20 - Artificial Intelligence II [ID:9329]
50 von 1026 angezeigt

The following content has been provided by the University

of Aelang and Nuremberg.

Our big topic now is machine learning.

And we're gearing up to neural networks.

And we have been looking at linear regression

and then linear

traditions.

looking at linear regression and then at neural networks, which basically use linear regression

in each unit.

And we'll see what that does.

So linear regression as essentially a learning method where we're learning classifications

from classifiers from examples.

We look at the regression problem first, so not classification, but regression, which

is essentially fit a linear model to a set of examples.

And we, as always, do the simple case first, which is univariate linear regression, meaning

a linear function of one real variable and one real value.

Very simple.

And that really means we have points in the xy plane, and we somehow have to find a linear

model that somehow describes that.

The question is, how do we do that?

And the answer is, just like always, we minimize the error in the hypothesis space of all linear

functions.

It's kind of always our approach right now.

And in this particular hypothesis space, there's a couple of things we used.

One is that linear models are determined by two values, the offset and the slope.

So we always talk about vectors w, which have offset and slope components.

And that gives us a linear function like this one.

And with that, we can then do things like minimize squared error loss, which is a nice

and analytic thing, which means we can look at partial derivatives.

And in this case, we can even minimize the loss and solve the equation here.

And what we've learned before and which we've used yesterday was the fact that this basic

idea somehow I think I should get rid of this here.

OK, so when we have this basic idea of minimizing the squared error loss, we can kind of attach

bells and whistles to this idea.

One of them was regularization.

Another might be to have some other kind of loss function and so on.

We could have the idea, well, but certain functions we want to avoid because we've looked

at them.

They're bad, so let's not go there or something like this.

So the idea is that even though in this case we can have a closed form solution, we should

kind of have a backup plan that still works when we've added bells and whistles.

And that is essentially gradient descent, which works extremely well here because in

this case, we have a convex function which has unique global minima.

OK, so in these cases, gradient descent works well.

And gradient descent always works the same way, is we basically loop over something.

We have a current point.

We have an update equation.

And we just iterate the update equation.

And if we're lucky, we have convergence relatively fast.

We often also introduce a learning rate parameter alpha here, which is typically smaller than

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

01:28:00 Min

Aufnahmedatum

2018-06-21

Hochgeladen am

2018-06-21 16:12:43

Sprache

en-US

Dieser Kurs beschäftigt sich mit den Grundlagen der Künstlichen Intelligenz (KI), insbesondere mit Techniken des Schliessens unter Unsicherheit, des maschinellen Lernens und dem Sprachverstehen. 
Der Kurs baut auf der Vorlesung Künstliche Intelligenz I vom Wintersemester auf und führt diese weiter. 

Lernziele und Kompetenzen
Fach- Lern- bzw. Methodenkompetenz

  • Wissen: Die Studierenden lernen grundlegende Repräsentationsformalismen und Algorithmen der Künstlichen Intelligenz kennen.

  • Anwenden: Die Konzepte werden an Beispielen aus der realen Welt angewandt (bungsaufgaben).

  • Analyse: Die Studierenden lernen über die Modellierung in der Maschine menschliche Intelligenzleistungen besser einzuschätzen.

Einbetten
Wordpress FAU Plugin
iFrame
Teilen